Liquid-liquid equilibrium data prediction using large margin nearest neighbor
Authors
Abstract:
Guanidine hydrochloride has been widely used in the initial recovery steps of active protein from the inclusion bodies in aqueous two-phase system (ATPS). The knowledge of the guanidine hydrochloride effects on the liquid-liquid equilibrium (LLE) phase diagram behavior is still inadequate and no comprehensive theory exists for the prediction of the experimental trends. Therefore the effect the guanidine hydrochloride on the phase behavior of PEG4000+ potassium phosphate+ water system at different guanidine hydrochloride concentrations and pH was investigated in this study. To fill the theoretical gaps, the typical of support vector machines was applied to the k-nearest neighbor method in order to develop a regression model to predict the LLE equilibrium of guanidine hydrochloride in the above mentioned system. Its advantage is its simplicity and good performance, with the disadvantage of an increase the execution time. The results of our method are quite promising: they were clearly better than those obtained by well-established methods such as Support Vector Machines, k-Nearest Neighbour and Random Forest. It is shown that the obtained results are more adequate than those provided by other common machine learning algorithms.
similar resources
Liquid-Liquid Equilibrium Data Prediction using Large Margin Nearest Neighbor
1Faculty of Chemical Engineering, Babol University of Technology, PO Box, 484, Babol, Iran 2Department of Chemical Engineering, Faculty of Chemical Engineering and Environmental Protection,“Gheorghe Asachi” Technical University of Iaşi, Str. Prof. Dr. Doc. DimitrieMangeron, nr. 73, 700050, Iaşi, Romania 3Department of Computer Science and Engineering, Faculty of Automatic Control and Computer E...
full textMixtures of Large Margin Nearest Neighbor Classifiers
The accuracy of the k-nearest neighbor algorithm depends on the distance function used to measure similarity between instances. Methods have been proposed in the literature to learn a good distance function from a labelled training set. One such method is the large margin nearest neighbor classifier that learns a global Mahalanobis distance. We propose a mixture of such classifiers where a gati...
full textLarge Margin Nearest Neighbor Classification using Curved Mahalanobis Distances
We consider the supervised classification problem of machine learning in Cayley-Klein projective geometries: We show how to learn a curved Mahalanobis metric distance corresponding to either the hyperbolic geometry or the elliptic geometry using the Large Margin Nearest Neighbor (LMNN) framework. We report on our experimental results, and further consider the case of learning a mixed curved Mah...
full textDrought Monitoring and Prediction using K-Nearest Neighbor Algorithm
Drought is a climate phenomenon which might occur in any climate condition and all regions on the earth. Effective drought management depends on the application of appropriate drought indices. Drought indices are variables which are used to detect and characterize drought conditions. In this study, it was tried to predict drought occurrence, based on the standard precipitation index (SPI), usin...
full textConvergence of Multi-pass Large Margin Nearest Neighbor Metric Learning
Large margin nearest neighbor classification (LMNN) is a popular technique to learn a metric that improves the accuracy of a simple knearest neighbor classifier via a convex optimization scheme. However, the optimization problem is convex only under the assumption that the nearest neighbors within classes remain constant. In this contribution we show that an iterated LMNN scheme (multi-pass LMN...
full textDistance Metric Learning for Large Margin Nearest Neighbor Classification
We show how to learn aMahanalobis distance metric for k-nearest neighbor (kNN) classification by semidefinite programming. The metric is trained with the goal that the k-nearest neighbors always belong to the same class while examples from different classes are separated by a large margin. On seven data sets of varying size and difficulty, we find that metrics trained in this way lead to signif...
full textMy Resources
Journal title
volume 13 issue 4
pages 14- 32
publication date 2016-11-01
By following a journal you will be notified via email when a new issue of this journal is published.
Hosted on Doprax cloud platform doprax.com
copyright © 2015-2023